RLAttn: An actor-critic model of eye movements during category learning

نویسندگان

  • Jordan I. Barnes
  • Caitlyn McColeman
  • Ekaterina R. Stepanova
  • Mark R. Blair
  • R. Calen Walshe
چکیده

Here we introduce a simple actor-critic model of eye movements during category learning that we call RLAttn (Reinforcement Learning of Attention). RLAttn stores the rewards it receives for making decisions or performing actions, while attempting to associate stimuli with particular categories. Over multiple trials, RLAttn learns that a large reward is most likely when the values of the relevant stimulus features have been revealed by fixations to them. The model is able to approximate human learning curves in a common category structure while generating fixation patterns similar to those found in human eye tracking data. We additionally observed that the model reduces its fixation counts to irrelevant features over the course of learning. We conclude with a discussion on the effective role eye movements might play in bridging structural credit assignment and temporal credit assignment problems.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

G Uide a Ctor - C Ritic for C Ontinuous C Ontrol

Actor-critic methods solve reinforcement learning problems by updating a parameterized policy known as an actor in a direction that increases an estimate of the expected return known as a critic. However, existing actor-critic methods only use values or gradients of the critic to update the policy parameter. In this paper, we propose a novel actor-critic method called the guide actor-critic (GA...

متن کامل

An Actor-critic Algorithm for Learning Rate Learning

Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks. It is observed that the models trained by SGD are sensitive to learning rates and good learning rates are problem specific. To avoid manually searching of learning rates, whic...

متن کامل

Actor-Critic Control with Reference Model Learning

We propose a new actor-critic algorithm for reinforcement learning. The algorithm does not use an explicit actor, but learns a reference model which represents a desired behaviour, along which the process is to be controlled by using the inverse of a learned process model. The algorithm uses Local Linear Regression (LLR) to learn approximations of all the functions involved. The online learning...

متن کامل

Reinforcement Learning for Learning Rate Control

Stochastic gradient descent (SGD), which updates the model parameters by adding a local gradient times a learning rate at each step, is widely used in model training of machine learning algorithms such as neural networks. It is observed that the models trained by SGD are sensitive to learning rates and good learning rates are problem specific. We propose an algorithm to automatically learn lear...

متن کامل

OnActor-Critic Algorithms

In this article, we propose and analyze a class of actor-critic algorithms. These are two-time-scale algorithms in which the critic uses temporal difference learning with a linearly parameterized approximation architecture, and the actor is updated in an approximate gradient direction, based on information provided by the critic. We show that the features for the critic should ideally span a su...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014